General vector space
Beyond Two Dimensions
Section 4.1: Vector Spaces and Subspaces
- Core Idea: This section establishes the fundamental "universe" or type of structure (a Vector Space) where the concepts of linear algebra primarily operate. It then defines well-behaved subsets (Subspaces) within these universes.
- Rationale for Vector Space (Definition 4.1.1):
- Abstraction and Generality: The goal is to identify the essential algebraic properties shared by many different mathematical settings where you can meaningfully "add" things and "scale" them by real numbers. Examples include familiar arrows in
, but also functions, matrices, sequences, and even more abstract examples. By defining an abstract structure based on common rules (the 10 axioms), we can prove theorems once for all vector spaces, rather than re-proving similar results for , then for functions, then for matrices, etc. This reveals deep connections and saves work. - Focus on Operations: The definition emphasizes the operations (addition and scalar multiplication) and the rules they follow (the axioms like commutativity, associativity, distributivity, existence of identities and inverses), rather than the specific nature of the objects ("vectors") themselves. A vector space is fundamentally a set where linear combinations make sense and behave nicely.
- Abstraction and Generality: The goal is to identify the essential algebraic properties shared by many different mathematical settings where you can meaningfully "add" things and "scale" them by real numbers. Examples include familiar arrows in
- Rationale for Subspace (Definition 4.1.12):
- Identifying Vector Spaces Within Vector Spaces: Often, we are interested in a subset
of a known vector space (e.g., a plane within , or continuous functions within all functions ). The question is: when does this subset , using the same addition and scalar multiplication inherited from , also qualify as a vector space in its own right? - Inherited Axioms vs. Closure: Many axioms (like commutativity) are automatically inherited by
because they hold true in the larger space . The crucial properties that might fail are: - Closure under Addition: Is
guaranteed to be back in whenever ? - Closure under Scalar Multiplication: Is
guaranteed to be back in whenever and ? - Contains Zero Vector: Does
contain the additive identity from ?
- Closure under Addition: Is
- The Definition: A subspace is defined precisely by these three essential properties. If a subset
satisfies these, all other vector space axioms will hold automatically (including the existence of additive inverses within ). - Significance: Subspaces allow us to study "flat" objects (like lines and planes through the origin in
) and other structured subsets (like continuous functions) using the tools of linear algebra. Proposition 4.1.16 is key because it shows that the span of any set of vectors is always a subspace, giving a primary method for constructing subspaces.
- Identifying Vector Spaces Within Vector Spaces: Often, we are interested in a subset
Let
- For all
, we have - For all
, if , then
Suppose that
- If
, then - If
, then
Let
- If
and for all , then , i.e. is the only additive identity - For all
, there is a unique with , i.e. additive inverses are unique
Let
3.
4.
5.
Let
Section 4.2: Solving Linear Systems
- Core Idea: This section develops the fundamental algorithm, Gaussian Elimination, for analyzing and solving systems of linear equations.
- Rationale for Focusing on Linear Systems:
- Central Tool: Many core questions in linear algebra (Is
in ? Is linearly independent? What are the coordinates ?) boil down to solving a system of linear equations. Having a robust method to solve these systems is therefore essential. - Structure and Algorithms: Linear systems possess a highly regular structure involving only sums of variables multiplied by constants. This regularity allows for the development of systematic, step-by-step algorithms like Gaussian elimination.
- Central Tool: Many core questions in linear algebra (Is
- Rationale for Augmented Matrix (Definition 4.2.7):
- Conciseness: Writing out variables (
) repeatedly is tedious and unnecessary. The coefficients and the constant terms on the right-hand side contain all the vital information. - Organization: The matrix format
neatly organizes this information and allows row operations to be performed systematically on the numbers without rewriting variables.
- Conciseness: Writing out variables (
- Rationale for Elementary Row Operations (Definition 4.2.3):
- Goal: To transform a complicated system into a simpler, equivalent one (having the same solution set).
- Why These Three?
- They are reversible (Prop 4.2.4). This is critical because it guarantees that we don't change the solution set when applying them (Corollary 4.2.5). We can always get back to the original system.
- They are sufficient to simplify any system into a standard "staircase" form (echelon form) where the solutions become easier to see.
- Rationale for Echelon Form (Definition 4.2.9 & 4.2.13):
- Systematic Goal: Gaussian elimination aims to reach an echelon form. This form makes the structure of the solutions apparent.
- Easy Solving: From echelon form, solutions can be found easily via back-substitution.
- Reduced Echelon Form: Going further to reduced echelon form (where leading entries are 1 and are the only non-zero entry in their column) makes the solutions even more explicit, often requiring little to no algebra during back-substitution.
- Rationale for Proposition 4.2.12 (Interpreting Echelon Form):
- Connecting Form to Solutions: This proposition directly links the visual pattern of the echelon form of the augmented matrix to the nature of the solution set (no solutions, unique solution, or infinitely many solutions) without needing to fully solve via back-substitution.
- Inconsistency: A leading entry in the augmented column (the rightmost one) signals an impossible equation like
where . - Consistency & Free Variables: If there's no inconsistency, the existence of columns without leading entries (corresponding to variables not solved for at the start of back-substitution) indicates the presence of free variables, which can be parameterized to generate infinite solutions. If every variable column has a leading entry, there are no free variables, yielding a unique solution.
In essence, Section 4.1 defines the abstract playing field (Vector Spaces) and important regions within it (Subspaces). Section 4.2 provides the essential computational toolkit (Gaussian Elimination on matrices) for answering concrete questions that arise when working within these abstract structures.
Suppose that we have a linear system, and we apply one elementary row operation to obtain a new system. We can then apply one elementary row operation to the new system in order to recover the old system.
Suppose that we have a linear system
For any matrix
Suppose that we have a linear system with augmented matrix
- If the last column of
contains a leading entry, then the system is inconsistent - If the last column of
contains no leading entry, but every other column of has a leading entry, then the system is consistent and has a unique solution - If the last column of
contains no leading entry, and there is at least one other column of without a leading entry, then the system is consistent and has infinitely many solutions. Moreover, for each choice of values for the variables that do not correspond to leading entries, there is a unique solution for the system taking these values.
Let
Section 4.3: Spanning Sequences
Let
If
If
Section 4.4: Linear Independence
Core Question of Section 4.4:
Section 4.3 asked: "Can we reach everything in the space?" Section 4.4 tackles the complementary question: "Do we have redundancy in our list of vectors?" In other words, is there a vector in our sequence
The Initial Idea of Redundancy:
The most direct way to think about redundancy is to ask if any vector
The Problem with the Initial Idea:
While intuitive, checking this directly is computationally inefficient. For a sequence of
The Rationale - A More Elegant Perspective:
- Rewriting Redundancy: Suppose a vector
is redundant. This means we can write . - The Homogeneous Equation: By simply moving
to the other side, we get:
.
Notice this is a linear combination of all the vectorsthat equals the zero vector . Crucially, at least one coefficient (the in front of ) is nonzero. - Linear Dependence: This motivates the definition. A sequence
is called linearly dependent if there exist scalars , not all zero, such that . This condition signals redundancy. - Linear Independence: Conversely, the sequence is linearly independent if the only way to make the linear combination
is by choosing the trivial solution where . This means there's no redundancy – no vector can be expressed in terms of the others (Proposition 4.4.2 proves this equivalence formally).
The Rationale - Connecting Linear Independence to Linear Systems:
How do we check if the only solution to
- Homogeneous System: This is equivalent to asking if the homogeneous linear system
has only the trivial solution , where is the matrix whose columns are . - Gaussian Elimination Again: We use Gaussian elimination on the coefficient matrix
(we can ignore the augmented column of zeros as it never changes). Let be an echelon form of . - Condition for Trivial Solution Only: The homogeneous system
(which has the same solution set as ) has only the trivial solution if and only if there are no free variables. - The Key Insight (Proposition 4.4.5): This happens exactly when the echelon form
has a leading entry (pivot) in every column. If there's a column without a pivot, that corresponds to a free variable, allowing for non-trivial solutions (meaning linear dependence). If every column has a pivot, there are no free variables, forcing the only solution to be the trivial one (meaning linear independence).
Rationale behind the Corollaries:
- Corollary 4.4.6: If
in is linearly independent, then . - Rationale: If the sequence is LI, the echelon form
of must have a pivot in every column (by Prop 4.4.5). There are columns, so there are pivots. Since each row can contain at most one pivot, you need at least rows to accommodate pivots. Thus, (the number of rows) must be . Intuitively, you can't have more independent "directions" (vectors) than the number of dimensions ( ) of the space they live in.
- Rationale: If the sequence is LI, the echelon form
- Corollary 4.4.7: If
, then in is linearly dependent. - Rationale: This is the contrapositive of Cor 4.4.6. If you have more vectors (
) than the dimension of the space ( ), they must be linearly dependent. An matrix with cannot have a pivot in every column (since there are at most pivots total, one per row), so there must be free variables when solving , guaranteeing non-trivial solutions. Intuitively, if you try to cram too many vectors into a lower-dimensional space, some must become redundant combinations of others.
- Rationale: This is the contrapositive of Cor 4.4.6. If you have more vectors (
In summary, Section 4.4 defines linear independence as the absence of non-trivial relationships between vectors (meaning no redundancy). It connects this concept to checking if a homogeneous system
Let
is linearly independent - There does not exist an
such that
Let
Let
If
If
Section 4.5: Bases and Coordinates
Core Idea of Section 4.5: Finding the "Right" Set of Vectors
The goal is to identify the most "efficient" set of vectors to describe a vector space
- Sufficiency (Spanning): The set must be large enough so that linear combinations of its vectors can create every vector in the space
. This is the idea of a spanning sequence from Section 4.3. We need . - No Redundancy (Linear Independence): The set should be "minimal" in the sense that it doesn't contain unnecessary vectors. If we can create one vector in the set using the others, it's redundant. This is precisely the idea captured by linear independence from Section 4.4. We need
to be linearly independent.
Rationale for "Basis" (Definition 4.5.1)
A basis is defined as a sequence of vectors
- It spans the entire space
. - It is linearly independent.
- Rationale: A basis hits the "sweet spot". It contains enough vectors to generate everything (it spans
), but it's lean enough that it contains no redundant information (it's linearly independent). It's the smallest possible set of vectors that can still generate the whole space.
Rationale for Theorem 4.5.2 (Basis
This theorem provides the crucial insight into why bases are so fundamental:
- Statement: A sequence
is a basis for if and only if every vector can be written as a linear combination in exactly one way (i.e., the scalars are unique for each ). - Rationale:
- The spanning property guarantees existence – that for any
, there is at least one way to write it as a linear combination. - The linear independence property guarantees uniqueness. If there were two different ways to write
, subtracting them would give a non-trivial linear combination of the basis vectors that equals , which contradicts linear independence. - Significance: This uniqueness is powerful. It means that once we fix a basis
, we can unambiguously associate any vector in our (potentially abstract) vector space with a specific list of real numbers .
- The spanning property guarantees existence – that for any
Rationale for "Coordinates" (Definition 4.5.3)
Since Theorem 4.5.2 guarantees a unique list of scalars
- Definition: The coordinates of
relative to is the vector in formed by these unique scalars: . - Rationale: The
function acts as a bridge or translator. It maps vectors in the potentially abstract space to concrete vectors in . This allows us to translate problems about into problems about , where we can use familiar tools like matrices and Gaussian elimination. It essentially imposes a coordinate system (like graph paper) onto , using the basis vectors as the axes.
Rationale for Propositions 4.5.5 and 4.5.6 (Constructing Bases)
These results provide ways to find a basis:
- Proposition 4.5.5: Any finite sequence that spans
can be "trimmed down" by removing redundant vectors until a basis is obtained. - Rationale: If a spanning set is linearly dependent, it contains redundancy (Prop 4.4.2). We can remove a redundant vector without changing the span (Prop 4.5.4). We repeat this until the set becomes linearly independent, at which point it's a basis because it still spans.
- Proposition 4.5.6: Provides a specific algorithm for doing this trimming process for vectors in
whose span is . It uses Gaussian elimination to identify the redundant columns (those without pivots) and tells us to keep the original columns that correspond to pivot columns in the echelon form. - Rationale: This works because row operations preserve the linear dependency relationships among the columns. The pivot columns of the echelon form are clearly linearly independent, and all other columns are combinations of them. This structure reflects the dependencies in the original columns.
In short, Section 4.5 defines the crucial concept of a basis as the optimal blend of spanning and linear independence. The core rationale is that a basis allows every vector in the space to be represented uniquely by a set of coordinates, effectively translating abstract vector space problems into the concrete setting of
Let
Let
Let
Let
Section 4.6: Dimension
Core Idea of Section 4.6: Defining the "Size" of a Vector Space
After establishing bases (Section 4.5) as "efficient" sets that both span the space and have no redundancy (linear independence), a natural question arises: Does every basis for a given vector space
The Rationale - Why Dimension is Well-Defined:
- The Problem: Vector spaces can have many different bases. For the concept of "dimension" to be meaningful as an intrinsic property of the space itself, we need to know that the number of vectors in a basis doesn't depend on which basis we happen to choose.
- Key Result (Theorem 4.6.2): This theorem is the cornerstone. It states that if a vector space
can be spanned by vectors , then any sequence with more than vectors (i.e., ) must be linearly dependent. - Rationale behind Thm 4.6.2: The proof uses the Steinitz Exchange Lemma (Proposition 4.6.1) repeatedly. The idea is that you can systematically replace vectors in the spanning set
with vectors from the linearly independent set while maintaining the span. Since you start with only vectors spanning the space, you can swap in at most of the 's. If you have more than vectors in the sequence ( ), then by the time you consider , it must be expressible as a linear combination of the preceding (which now span the space), proving the sequence is linearly dependent.
- Rationale behind Thm 4.6.2: The proof uses the Steinitz Exchange Lemma (Proposition 4.6.1) repeatedly. The idea is that you can systematically replace vectors in the spanning set
- Consequence (Corollary 4.6.3): Taking the contrapositive of Theorem 4.6.2, we get: If a space
is spanned by and is linearly independent in , then . In words: the size of any linearly independent set is less than or equal to the size of any spanning set. - Invariance of Basis Size (Corollary 4.6.4): This directly uses the previous corollary. If you have two bases,
with vectors and with vectors: - Since
spans and is linearly independent, . - Since
spans and is linearly independent, . - Therefore,
.
- Since
- Definition of Dimension (Definition 4.6.5): Because Corollary 4.6.4 guarantees that all bases for a given vector space
(if any exist) have the same number of elements, we can unambiguously define the dimension of , denoted , as this common number. - Rationale: Dimension captures the intrinsic "size" or number of "degrees of freedom" or independent "directions" within the vector space, regardless of the specific basis chosen to measure it.
The Rationale - Consequences and Uses of Dimension:
Once dimension is well-defined, it becomes a powerful tool:
- Basis Shortcut (Proposition 4.6.7): If you know
, then checking if a sequence of exactly vectors is a basis becomes easier. You only need to check one of the two conditions: - If it's linearly independent, it must also span
. - If it spans
, it must also be linearly independent. - Rationale: This follows from Theorem 4.6.2 and its corollary. For instance, if you have
linearly independent vectors, but they didn't span , you could add another vector to get linearly independent vectors (Prop 4.6.9), which contradicts the fact that the -dimensional space cannot contain linearly independent vectors (Cor 4.6.3). A similar argument works the other way.
- If it's linearly independent, it must also span
- Subspace Dimension (Proposition 4.6.8): Any subspace
of a finite-dimensional vector space is also finite-dimensional, and . - Rationale: You can construct a basis for
by starting with a linearly independent set in and extending it (using Prop 4.6.10, below). This process must stop because is contained in , and you can't have more linearly independent vectors in than the dimension of . This guarantees has a finite basis, and the number of vectors must be less than or equal to .
- Rationale: You can construct a basis for
- Extending Linearly Independent Sets (Proposition 4.6.10): Any linearly independent sequence in a finite-dimensional vector space can be extended to form a basis for that space.
- Rationale: Start with the linearly independent set. If it doesn't already span the whole space, find a vector not in its span and add it to the set; the new set remains linearly independent (Prop 4.6.9). Repeat this process. Since the dimension is finite, you can't keep adding vectors indefinitely while maintaining linear independence (by Thm 4.6.2). The process must stop when the set spans the space, at which point it is a basis. This confirms that linearly independent sets act as valid starting points or "skeletons" for building bases.
In summary, Section 4.6 leverages the interplay between spanning sets and linearly independent sets (specifically Theorem 4.6.2) to rigorously establish that the size of a basis is an invariant property of a vector space, allowing the definition of dimension. Dimension then becomes a key characteristic used to understand the structure of the space, relationships between spaces and subspaces, and provides useful criteria for determining if a set of vectors forms a basis.
Let
Let
Let
Suppose that
Let
- If
is linearly independent, then is a basis of - If
, then is a basis of
Let
Let
is linearly independent
Let